home
***
CD-ROM
|
disk
|
FTP
|
other
***
search
/
EnigmA Amiga Run 1995 November
/
EnigmA AMIGA RUN 02 (1995)(G.R. Edizioni)(IT)[!][issue 1995-11][Skylink CD].iso
/
earcd
/
docs
/
announc5.lha
/
1000
/
1041.edit
< prev
next >
Wrap
Internet Message Format
|
1995-01-28
|
4KB
From: JamesB@werple.mira.net.au (James Burton)
Newsgroups: comp.sys.amiga.announce
Path: toadflax.cs!not-for-mail
Subject: GetURL 1.0 Web page downloader available for FTP
Reply-To: JamesB@werple.mira.net.au
Followup-To: comp.sys.amiga.networking
Organization: Melbourne ArtWorks
Message-Number: 1041
Approved: zerkle@cs.ucdavis.edu
TITLE
GetURL.rexx
VERSION
1.0
AUTHOR
James Burton
c/o
Department of Computer Science & Computer Engineering
Latrobe University
Bundoora, Victoria, 3083
Australia
EMail: burton@cs.latrobe.edu.au
Web: http://www.cs.latrobe.edu.au/~burton/
DESCRIPTION
-- Script to download HTML systems across the network --
GetURL.rexx is an ARexx script which will download World-Wide Web pages.
With a simple command line it will download a specific page, and with
more complex command lines it will be able to download specific sets of
documents.
The intention was to create a tool that allowed local caching of important
Web pages and a flexible way of specifying what pages are important. The
script has no GUI as of yet but may have at some stage in the future.
If you have ever tried to download and save to disc a 200 page document
using Mosaic, then you know what this script is for. Mosaic will only
let you load a page, then load it to disc, then load another page etc.
This is a very frustrating process. GetURL automates this process and
will run in batch mode without user intervention.
The major features of GetURL.rexx are as follows:
* doesn't require AMosaic, you you can be browsing something else
with AMosaic whilst this is running
* save pages to your hard disc so that they can be read offline and
you can also give them to friends on a floppy disc. Who know,
you may even be able to sell discs containing web pages :-)
* flexible set of command line switches that allow you to restrict the
type of pages that it downloads
* ability to specify files for the lists of URLs that it keeps so
that any search for pages can be stopped and restarted at a later
date. i.e. you could run GetURL for 2 hours a day whilst you are
online and gradually download everything in the entire universe
and it won't repeat itself.
* ability to download itself when there are new versions.
* will use a proxy if you have access to one, in order to both speed up
access to pages and also to reduce network load.
* will download binary files (*.gif, *.lha) as easily as text and html
files.
* documentation is in the top of the script file.
SPECIAL REQUIREMENTS
* Until somebody writes a TCP: device for the AS225 TCP/IP protocol
stack unfortunately this script will require AmiTCP.
* requires the TCP: device be mounted
* either restraint, or an extremely large hard disc - your choice :-)
* This script is no use at all unless you have AmiTCP set up and running.
If you don't know what this means then please ask me
(burton@cs.latrobe.edu.au).
LOCATION
This script is available via
Anonymous FTP to AmiNet
wuarchive.wustl.edu (128.252.135.4)
/pub/aminet/comm/tcp/GetURL-1.0.rexx
and all of it's mirrors. Please check the closest mirror FIRST.
HTTP from my university account
http://www.cs.latrobe.edu.au/~burton/Public/GetURL.rexx
(this URL will always point to the newest version)
FILE NAMES
GetURL-1.0.rexx 41770 Bytes
GetURL-1.0.readme 3874 Bytes
PRICE
Absolutely free to humans
DISTRIBUTABILITY
Public domain. But so that a hundred different versions of this don't
appear, please send corrections, new features, bug fixes etc. to me
and I will coordinate.
--
James Burton |
EMail: JamesB@werple.mira.net.au | Latrobe University
WWW : http://www.cs.latrobe.edu.au/~burton/ | Melbourne, Australia